A Video Synthesizer is a device that electronically creates a video signal.
A video synthesizer is able to generate a variety of visual material without camera input through the use of internal video pattern generators, as seen in the stillframes of motion sequences shown above. It can also accept and "clean up and enhance" or "distort" live television camera imagery. The synthesizer creates a wide range of imagery through purely electronic manipulations. This imagery is visible within the output video signal when this signal is displayed. The output video signal can be viewed on a wide range of conventional video equipment, such as TV monitors, theater video projectors, computer displays, etc.
Video pattern generators may produce static or moving or evolving imagery. Examples include geometric patterns ( in 2D or 3D ), subtitle text characters in a particular font, or weather maps.
Imagery from TV cameras can be altered in color or geometrically scaled, tilted, wrapped around objects, and otherwise manipulated.
A particular video synthesizer will offer a subset of possible effects.
Contents |
The history of video synthesis is tied in to a "real time performance" ethic. The equipment is usually expected to function on input camera signals the machine has never seen before, delivering a processed signal continuously and with a minimum of delay in response to the ever changing live video inputs. Following in the tradition of performance instruments of the audio synthesis world such as the Theremin, video synthesizers were designed with the expectation they would be played in live concert theatrical situations or set up in a studio ready to process a videotape from a playback VCR in real time while recording the results on a second VCR. Venues of these performances included "Electronic Visualization Events" in Chicago, The Kitchen in NYC, and museum installations. Video artist/performer Don Slepian [1] designed, built and performed a foot-controlled Visual Instrument at the Centre Pompideau in Paris (1983) and the NY Open Center that combined genlocked Genlocking early micro-computers Apple II Plus with the Chromaton 14 Video Synthesizer [2] and channels of colorized video feedback.
Analog and early real time digital synthesizers existed before modern computer 3D modeling. Typical 3D renderers are not real time, as they concentrate on computing each frame from, for example, a recursive ray tracing algorithm, however long it takes. This distinguishes them from video synthesizers, which MUST deliver a new output frame by the time the last one has been shown, and repeat this performance continuously ( typically delivering a new frame regularly every 1/30 or 1/25 of a second ) . The real time constraint results in a difference in design philosophy between these two classes of systems.
Video synthesizers overlap with video special effects equipment used in real time network television broadcast and post-production situations. Many innovations in television broadcast equipment as well as computer graphics displays evolved from synthesizers developed in the video artists' community and these industries often support "electronic art projects" in this area to show appreciation of this history.
Many principles used in the construction of early video synthesizers reflected a healthy and dynamic interplay between electronic requirements and traditional interpretations of artistic forms. For example, Rutt & Etra and Sandin carried forward as an essential principle ideas of Robert Moog that standardized signal ranges so that any module's output could be connected to "voltage control" any other module's input. The consequence of this in a machine like the Rutt-Etra was that position, brightness, and color were completely interchangeable and could be used to modulate each other during the processing that led to the final image. Videotapes by Louise and Bill Etra and Steina and Woody Vasulka dramatized this new class of effects. This led to various interpretations of the multi-modal synthesesia of these aspects of the image in dialogues that extended the McLuhanesque language of film criticism of the time.
In the UK Richard Monkhouse working for EMS developed a hybrid video synthesiser - Spectre - later renamed 'Spectron' which used the EMS patchboard system to allow completely flexible connections between module inputs and outputs. The video signals were digital, but they were controlled by analog voltages. There was a digital patchboard for image composition and an analog patchboard for motion control.
Video synthesizers moved from analog to the precision control of digital. The first digital effects as exemplified by Stephen Beck's Video Weavings used digital oscillators optionally linked to horizontal, vertical, or frame resets to generate timing ramps. These ramps could be gated to create the video image itself and were responsible for its underlying geometric texture. Schier and Vasulka advanced the state of the art from address counters to programmable (microcodable) AMD Am2901 bit slice based address generators. On the data path, they used 74S181 arithmetic and logic units, previously thought of as a component for doing arithmetic instructions in minicomputers, to process real time video signals, creating new signals representing the sum, difference, AND, XOR, and so on, of two input signals. These two elements, the address generator, and the video data pipeline, recur as core features of digital video architecture.
The address generator supplied read and write addresses to a real time video memory, which can be thought of as evolution into the most flexible form of gating the address bits together to produce the video. While the video frame buffer is now present in every computer's graphics card, it has not carried forward a number of features of the early video synths. The address generator counts in a fixed rectangular pattern from the upper left hand corner of the screen, across each line, to the bottom. This discarded a whole technology of modifying the image by variations in the read and write addressing sequence provided by the hardware address generators as the image passed through the memory. Today, address based distortions are more often accomplished by blitter operations moving data in the memory, rather than changes in video hardware addressing patterns.
Books
Web
Don Slepian music videos: 1. "Beginnings", 1983 - Chromascope Analog Video Synthesizer [8] 2. "Next Time", 1983 - Chromascope with Luma-Keying and Image Processing [9] 3. "Rising Crimson Tide", 1982 - Chromaton 14 Analog Video Synthesizer with colorized video feedback techniques, animations from the Apple II+ micocomputer with genlock board. [10]